func runtime.alignDown
16 uses
runtime (current package)
malloc.go#L1688: setTinyBlockContext(unsafe.Pointer(alignDown(uintptr(x), maxTinySize)))
mbitmap.go#L308: tp.addr = tp.elem + alignDown(n-(tp.elem-oldelem), ptrBits*goarch.PtrSize)
mbitmap.go#L310: tp.addr += alignDown(n, ptrBits*goarch.PtrSize)
mbitmap.go#L799: size -= alignDown(uintptr(cheaprand())%size, goarch.PtrSize)
mcheckmark.go#L114: wordIdx := (obj - alignDown(obj, heapArenaBytes)) / goarch.PtrSize
mem_linux.go#L108: end := alignDown(uintptr(v)+n, physHugePageSize)
mgcscavenge.go#L976: hugePageBelow := uint(alignDown(uintptr(start), pagesPerHugePage))
mheap.go#L2407: ptr = alignDown(ptr, maxTinySize)
mpagealloc.go#L180: return int(alignDown(uintptr(lo), e)), int(alignUp(uintptr(hi), e))
mpagealloc.go#L366: base = alignDown(base, pallocChunkBytes)
mpagealloc_64bit.go#L118: baseOffset := alignDown(uintptr(sumIdxBase)*pallocSumBytes, physPageSize)
mpagealloc_64bit.go#L210: needMin := alignDown(uintptr(chunkIndex(base)), physPageSize/scSize)
mpagecache.go#L138: base: chunkBase(ci) + alignDown(uintptr(j), 64)*pageSize,
mpagecache.go#L155: base: alignDown(addr, 64*pageSize),
mranges.go#L89: limit := alignDown(a.limit.addr()-len, uintptr(align))
stubs.go#L372: func alignDown(n, a uintptr) uintptr {